首页> 外文OA文献 >Deep Reinforcement Learning with Successor Features for Navigation across Similar Environments
【2h】

Deep Reinforcement Learning with Successor Features for Navigation across Similar Environments

机译:深度强化学习与后续导航功能   跨类似的环境

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

In this paper we consider the problem of robot navigation in simple maze-likeenvironments where the robot has to rely on its onboard sensors to perform thenavigation task. In particular, we are interested in solutions to this problemthat do not require localization, mapping or planning. Additionally, we requirethat our solution can quickly adapt to new situations (e.g., changingnavigation goals and environments). To meet these criteria we frame thisproblem as a sequence of related reinforcement learning tasks. We propose asuccessor feature based deep reinforcement learning algorithm that can learn totransfer knowledge from previously mastered navigation tasks to new probleminstances. Our algorithm substantially decreases the required learning timeafter the first task instance has been solved, which makes it easily adaptableto changing environments. We validate our method in both simulated and realrobot experiments with a Robotino and compare it to a set of baseline methodsincluding classical planning-based navigation.
机译:在本文中,我们考虑了在迷宫般简单环境中的机器人导航问题,在这种情况下,机器人必须依靠其机载传感器来执行导航任务。特别是,我们对不需要本地化,映射或规划的解决方案感兴趣。此外,我们要求我们的解决方案能够快速适应新情况(例如,更改导航目标和环境)。为了满足这些标准,我们将此问题定义为一系列相关的强化学习任务。我们提出了基于后继特征的深度强化学习算法,该算法可以学习将知识从以前掌握的导航任务转移到新的问题实例。解决第一个任务实例后,我们的算法大大减少了所需的学习时间,这使其很容易适应不断变化的环境。我们使用Robotino在模拟和真实机器人实验中验证了我们的方法,并将其与一组基线方法(包括基于经典计划的导航)进行了比较。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号